18 research outputs found
Duality Mapping for Schatten Matrix Norms
In this paper, we fully characterize the duality mapping over the space of
matrices that are equipped with Schatten norms. Our approach is based on the
analysis of the saturation of the H\"older inequality for Schatten norms. We
prove in our main result that, for , the duality mapping over
the space of real-valued matrices with Schatten- norm is a continuous and
single-valued function and provide an explicit form for its computation. For
the special case , the mapping is set-valued; by adding a rank
constraint, we show that it can be reduced to a Borel-measurable single-valued
function for which we also provide a closed-form expression
Generating Sparse Stochastic Processes Using Matched Splines
We provide an algorithm to generate trajectories of sparse stochastic
processes that are solutions of linear ordinary differential equations driven
by L\'evy white noises. A recent paper showed that these processes are limits
in law of generalized compound-Poisson processes. Based on this result, we
derive an off-the-grid algorithm that generates arbitrarily close
approximations of the target process. Our method relies on a B-spline
representation of generalized compound-Poisson processes. We illustrate
numerically the validity of our approach
Linear Inverse Problems with Hessian-Schatten Total Variation
In this paper, we characterize the class of extremal points of the unit ball
of the Hessian-Schatten total variation (HTV) functional. The underlying
motivation for our work stems from a general representer theorem that
characterizes the solution set of regularized linear inverse problems in terms
of the extremal points of the regularization ball. Our analysis is mainly based
on studying the class of continuous and piecewise linear (CPWL) functions. In
particular, we show that in dimension , CPWL functions are dense in the
unit ball of the HTV functional. Moreover, we prove that a CPWL function is
extremal if and only if its Hessian is minimally supported. For the converse,
we prove that the density result (which we have only proven for dimension ) implies that the closure of the CPWL extreme points contains all extremal
points
Deep Neural Networks with Trainable Activations and Controlled Lipschitz Constant
We introduce a variational framework to learn the activation functions of
deep neural networks. Our aim is to increase the capacity of the network while
controlling an upper-bound of the actual Lipschitz constant of the input-output
relation. To that end, we first establish a global bound for the Lipschitz
constant of neural networks. Based on the obtained bound, we then formulate a
variational problem for learning activation functions. Our variational problem
is infinite-dimensional and is not computationally tractable. However, we prove
that there always exists a solution that has continuous and piecewise-linear
(linear-spline) activations. This reduces the original problem to a
finite-dimensional minimization where an l1 penalty on the parameters of the
activations favors the learning of sparse nonlinearities. We numerically
compare our scheme with standard ReLU network and its variations, PReLU and
LeakyReLU and we empirically demonstrate the practical aspects of our
framework
Deep Spline Networks With Control Of Lipschitz Regularity
The motivation for this work is to improve the performance of deep neural networks through the optimization of the individual activation functions. Since the latter results in an infinite-dimensional optimization problem, we resolve the ambiguity by searching for the sparsest and most regular solution in the sense of Lipschitz. To that end, we first introduce a bound that relates the properties of the pointwise nonlinearities to the global Lipschitz constant of the network. By using the proposed bound as regularizer, we then derive a representer theorem that shows that the optimum configuration is achievable by a deep spline network. It is a variant of a conventional deep ReLU network where each activation function is a piecewise-linear spline with adaptive knots. The practical interest is that the underlying spline activations can be expressed as linear combinations of ReLU units and optimized using l(1)-minimization techniques
Wavelet analysis of the Besov regularity of Levy white noise
We characterize the local smoothness and the asymptotic growth rate of the Levy white noise. We do so by characterizing the weighted Besov spaces in which it is located. We extend known results in two ways. First, we obtain new bounds for the local smoothness via the Blumenthal-Getoor indices of the Levy white noise. We also deduce the critical local smoothness when the two indices coincide, which is true for symmetric-a-stable, compound Poisson, and symmetric-gamma white noises to name a few. Second, we express the critical asymptotic growth rate in terms of the moment properties of the Levy white noise. Previous analyses only provided lower bounds for both the local smoothness and the asymptotic growth rate. Showing the sharpness of these bounds requires us to determine in which Besov spaces a given Levy white noise is (almost surely) not. Our methods are based on the wavelet-domain characterization of Besov spaces and precise moment estimates for the wavelet coefficients of the Levy white noise
The Wavelet Compressibility of Compound Poisson Processes
In this paper, we precisely quantify the wavelet compressibility of compound Poisson processes. To that end, we expand the given random process over the Haar wavelet basis and we analyse its asymptotic approximation properties. By only considering the nonzero wavelet coefficients up to a given scale, what we call the greedy approximation, we exploit the extreme sparsity of the wavelet expansion that derives from the piecewise-constant nature of compound Poisson processes. More precisely, we provide lower and upper bounds for the mean squared error of greedy approximation of compound Poisson processes. We are then able to deduce that the greedy approximation error has a sub-exponential and super-polynomial asymptotic behavior. Finally, we provide numerical experiments to highlight the remarkable ability of wavelet-based dictionaries in achieving highly compressible approximations of compound Poisson processes